Agnostically Learning Halfspaces with Margin Errors
نویسندگان
چکیده
We describe and analyze a new algorithm for agnostically learning half-spaces with respect to the margin error rate. Roughly speaking, this corre-sponds to the worst-case error rate after each point is perturbed by a noisevector of length at most μ. Margin based analysis is widely used in learningtheory and is considered the most successful theoretical explanation for thestatistical properties of several learning algorithms, such as Support VectorMachines and AdaBoost. The proposed algorithm can learn n-dimensionalhalfspaces in time poly(n exp( 1μ log( 1μ ))), for any distribution, where μ isthe margin parameter and the error rate of the learned classifier is at mostplus the margin error rate of the optimal halfspace. This improves overthe bound poly(n exp(( 1μ) 2 log( 1μ ))), derived by Ben-David and Simon [7].If the distribution over instances is uniform on the unit ball, we recoverthe poly(n1/ 4) complexity bound of Kalai et al [17]. Furthermore, the de-pendence on the dimension n in our complexity bound can be replaced bythe time required to calculate inner-products. This enables us to efficientlylearn halfspaces in possibly infinite dimensional Hilbert spaces by using theso-called kernel trick.
منابع مشابه
Learning Kernel-Based Halfspaces with the 0-1 Loss
We describe and analyze a new algorithm for agnostically learning kernel-based halfspaces with respect to the 0-1 loss function. Unlike most of the previous formulations, which rely on surrogate convex loss functions (e.g., hinge-loss in support vector machines (SVMs) and log-loss in logistic regression), we provide finite time/sample guarantees with respect to the more natural 0-1 loss functio...
متن کاملLearning Kernel-Based Halfspaces with the Zero-One Loss
We describe and analyze a new algorithm for agnostically learning kernel-based halfspaces with respect to the zero-one loss function. Unlike most previous formulations which rely on surrogate convex loss functions (e.g. hinge-loss in SVM and log-loss in logistic regression), we provide finite time/sample guarantees with respect to the more natural zero-one loss function. The proposed algorithm ...
متن کاملA PTAS for Agnostically Learning Halfspaces
We present a PTAS for agnostically learning halfspaces w.r.t. the uniform distribution on the d dimensional sphere. Namely, we show that for every μ > 0 there is an algorithm that runs in time poly ( d, 1 ) , and is guaranteed to return a classifier with error at most (1 + μ)opt + , where opt is the error of the best halfspace classifier. This improves on Awasthi, Balcan and Long [2] who showed...
متن کاملEmbedding Hard Learning Problems Into Gaussian Space
We give the first representation-independent hardness result for agnostically learning halfspaces with respect to the Gaussian distribution. We reduce from the problem of learning sparse parities with noise with respect to the uniform distribution on the hypercube (sparse LPN), a notoriously hard problem in theoretical computer science and show that any algorithm for agnostically learning halfs...
متن کاملPotential-Based Agnostic Boosting
We prove strong noise-tolerance properties of a potential-based boosting algorithm, similar to MadaBoost (Domingo and Watanabe, 2000) and SmoothBoost (Servedio, 2003). Our analysis is in the agnostic framework of Kearns, Schapire and Sellie (1994), giving polynomial-time guarantees in presence of arbitrary noise. A remarkable feature of our algorithm is that it can be implemented without reweig...
متن کامل